The oracle penalty method

نویسندگان

  • Martin Schlüter
  • Matthias Gerdts
چکیده

A new and universal penalty method is introduced in this contribution. It is especially intended to be applied in stochastic metaheuristics like genetic algorithms, particle swarm optimization or ant colony optimization. The novelty of this method is, that it is an advanced approach that only requires one parameter to be tuned. Moreover this parameter, named oracle, is easy and intuitive to handle. A pseudo-code implementation of the method is presented together with numerical results on a set of 60 constrained benchmark problems from the open literature. The results are compared with those obtained by common penalty methods, revealing the strength of the proposed approach. Further results on three real-world applications are briefly discussed and fortify the practical usefulness and capability of the method.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Asymptotic properties for combined L 1 and concave

Two important goals of high-dimensional modeling are prediction and variable selection. In this article, we consider regularization with combined L1 and concave penalties, and study the sampling properties of the global optimum of the suggested method in ultra-high dimensional 10 settings. The L1-penalty provides the minimum regularization needed for removing noise variables in order to achieve...

متن کامل

The Sparse Laplacian Shrinkage Estimator for High-Dimensional Regression.

We propose a new penalized method for variable selection and estimation that explicitly incorporates the correlation patterns among predictors. This method is based on a combination of the minimax concave penalty and Laplacian quadratic associated with a graph as the penalty function. We call it the sparse Laplacian shrinkage (SLS) method. The SLS uses the minimax concave penalty for encouragin...

متن کامل

Robust Regression through the Huber’s criterion and adaptive lasso penalty

The Huber’s Criterion is a useful method for robust regression. The adaptive least absolute shrinkage and selection operator (lasso) is a popular technique for simultaneous estimation and variable selection. The adaptive weights in the adaptive lasso allow to have the oracle properties. In this paper we propose to combine the Huber’s criterion and adaptive penalty as lasso. This regression tech...

متن کامل

Variable Selection for Cox’s Proportional Hazards Model and Frailty Model By

A class of variable selection procedures for parametric models via nonconcave penalized likelihood was proposed in Fan and Li (2001a). It has been shown there that the resulting procedures perform as well as if the subset of significant variables were known in advance. Such a property is called an oracle property. The proposed procedures were illustrated in the context of linear regression, rob...

متن کامل

Penalty methods with stochastic approximation for stochastic nonlinear programming

In this paper, we propose a class of penalty methods with stochastic approximation for solving stochastic nonlinear programming problems. We assume that only noisy gradients or function values of the objective function are available via calls to a stochastic first-order or zeroth-order oracle. In each iteration of the proposed methods, we minimize an exact penalty function which is nonsmooth an...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • J. Global Optimization

دوره 47  شماره 

صفحات  -

تاریخ انتشار 2010